Combining Nearest Neighbor Classi ers Through Multiple

نویسنده

  • Stephen D Bay
چکیده

Combining multiple classiiers is an eeective technique for improving accuracy. There are many general combining algorithms, such as Bagging or Error Correcting Output Coding, that signiicantly improve classiiers like decision trees, rule learners, or neural networks. Unfortunately, many combining methods do not improve the nearest neighbor classiier. In this paper, we present MFS, a combining algorithm designed to improve the accuracy of the nearest neighbor (NN) classiier. MFS combines multiple NN classiiers each using only a random subset of features. The experimental results are encouraging: On 25 datasets from the UCI Repository, MFS sig-niicantly improved upon the NN, k nearest neighbor (kNN), and NN classiiers with forward and backward selection of features. MFS was also robust to corruption by irrelevant features compared to the kNN classiier. Finally, we show that MFS is able to reduce both bias and variance components of error.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Combining Nearest Neighbor Classifiers Through Multiple Feature Subsets

Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, many combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a combining a...

متن کامل

Nearest Neighbor Classi cation from Multiple Feature Subsets

Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging, Boosting, or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, these combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a...

متن کامل

Eective supra-classi®ers for knowledge base construction

We explore the use of the supra-classi®er framework in the construction of a classi®er knowledge base. Previously, we introduced this framework within which labels produced by old classi®ers are used to improve the generalization performance of a new classi®er for a di€erent but related classi®cation task (Bollacker and Ghosh, 1998). We showed empirically that a simple Hamming nearest neighbor ...

متن کامل

Nearest neighbor classi®er: Simultaneous editing and feature selection

Nearest neighbor classi®ers demand signi®cant computational resources (time and memory). Editing of the reference set and feature selection are two di€erent approaches to this problem. Here we encode the two approaches within the same genetic algorithm (GA) and simultaneously select features and reference cases. Two data sets were used: the SATIMAGE data and a generated data set. The GA was fou...

متن کامل

Cloud Classi cation Using Error-Correcting Output Codes

Novel arti cial intelligence methods are used to classify 16x16 pixel regions (obtained from Advanced Very High Resolution Radiometer (AVHRR) images) in terms of cloud type (e.g., stratus, cumulus, etc.). We previously reported that intelligent feature selection methods, combined with nearest neighbor classi ers, can dramatically improve classi cation accuracy on this task. Our subsequent analy...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1998